Improving Lasso for model selection and prediction

نویسندگان

چکیده

It is known that the Thresholded Lasso (TL), SCAD or MCP correct intrinsic estimation bias of Lasso. In this paper we propose an alternative method improving for predictive models with general convex loss functions which encompass normal linear models, logistic regression, quantile support vector machines. For a given penalty order absolute values nonzero coefficients and then select final model from small nested family by Generalized Information Criterion. We derive exponential upper bounds on selection error method. These results confirm that, at least our algorithm seems to be benchmark theory as it constructive, computationally efficient leads consistent under weak assumptions. Constructivity means in contrast TL, MCP, does not rely unknown parameters cone invertibility factor. Instead, only needs sample size, number predictors bound noise parameter. show numerical experiments synthetic real-world datasets implementation more accurate than implementations studied concave regularizations. Our procedure included R package DMRnet available CRAN repository.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

On Model Selection Consistency of Lasso On Model Selection Consistency of Lasso

Sparsity or parsimony of statistical models is crucial for their proper interpretations, as in sciences and social sciences. Model selection is a commonly used method to find such models, but usually involves a computationally heavy combinatorial search. Lasso (Tibshirani, 1996) is now being used as a computationally feasible alternative to model selection. Therefore it is important to study La...

متن کامل

On Model Selection Consistency of Lasso

Sparsity or parsimony of statistical models is crucial for their proper interpretations, as in sciences and social sciences. Model selection is a commonly used method to find such models, but usually involves a computationally heavy combinatorial search. Lasso (Tibshirani, 1996) is now being used as a computationally feasible alternative to model selection. Therefore it is important to study La...

متن کامل

A Note on the Lasso for Gaussian Graphical Model Selection

Inspired by the success of the Lasso for regression analysis (Tibshirani, 1996), it seems attractive to estimate the graph of a multivariate normal distribution by `1-norm penalised likelihood maximisation. The objective function is convex and the graph estimator can thus be computed efficiently, even for very large graphs. However, we show in this note that the resulting estimator is not consi...

متن کامل

Improved Lasso for genomic selection.

Empirical experience with genomic selection in dairy cattle suggests that the distribution of the effects of single nucleotide polymorphisms (SNPs) might be far from normality for some traits. An alternative, avoiding the use of arbitrary prior information, is the Bayesian Lasso (BL). Regular BL uses a common variance parameter for residual and SNP effects (BL1Var). We propose here a BL with di...

متن کامل

The lasso method for variable selection in the Cox model.

I propose a new method for variable selection and shrinkage in Cox's proportional hazards model. My proposal minimizes the log partial likelihood subject to the sum of the absolute values of the parameters being bounded by a constant. Because of the nature of this constraint, it shrinks coefficients and produces some coefficients that are exactly zero. As a result it reduces the estimation vari...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Scandinavian Journal of Statistics

سال: 2021

ISSN: ['0303-6898', '1467-9469']

DOI: https://doi.org/10.1111/sjos.12546